171 research outputs found
DiffLLE: Diffusion-guided Domain Calibration for Unsupervised Low-light Image Enhancement
Existing unsupervised low-light image enhancement methods lack enough
effectiveness and generalization in practical applications. We suppose this is
because of the absence of explicit supervision and the inherent gap between
real-world scenarios and the training data domain. In this paper, we develop
Diffusion-based domain calibration to realize more robust and effective
unsupervised Low-Light Enhancement, called DiffLLE. Since the diffusion model
performs impressive denoising capability and has been trained on massive clean
images, we adopt it to bridge the gap between the real low-light domain and
training degradation domain, while providing efficient priors of real-world
content for unsupervised models. Specifically, we adopt a naive unsupervised
enhancement algorithm to realize preliminary restoration and design two
zero-shot plug-and-play modules based on diffusion model to improve
generalization and effectiveness. The Diffusion-guided Degradation Calibration
(DDC) module narrows the gap between real-world and training low-light
degradation through diffusion-based domain calibration and a lightness
enhancement curve, which makes the enhancement model perform robustly even in
sophisticated wild degradation. Due to the limited enhancement effect of the
unsupervised model, we further develop the Fine-grained Target domain
Distillation (FTD) module to find a more visual-friendly solution space. It
exploits the priors of the pre-trained diffusion model to generate
pseudo-references, which shrinks the preliminary restored results from a coarse
normal-light domain to a finer high-quality clean field, addressing the lack of
strong explicit supervision for unsupervised methods. Benefiting from these,
our approach even outperforms some supervised methods by using only a simple
unsupervised baseline. Extensive experiments demonstrate the superior
effectiveness of the proposed DiffLLE
A SIMPLE Approach to Provably Reconstruct Ising Model with Global Optimality
Reconstruction of interaction network between random events is a critical
problem arising from statistical physics and politics to sociology, biology,
and psychology, and beyond. The Ising model lays the foundation for this
reconstruction process, but finding the underlying Ising model from the least
amount of observed samples in a computationally efficient manner has been
historically challenging for half a century. By using the idea of sparsity
learning, we present a approach named SIMPLE that has a dominant sample
complexity from theoretical limit. Furthermore, a tuning-free algorithm is
developed to give a statistically consistent solution of SIMPLE in polynomial
time with high probability. On extensive benchmarked cases, the SIMPLE approach
provably reconstructs underlying Ising models with global optimality. The
application on the U.S. senators voting in the last six congresses reveals that
both the Republicans and Democrats noticeably assemble in each congresses;
interestingly, the assembling of Democrats is particularly pronounced in the
latest congress
Identifying Hard Noise in Long-Tailed Sample Distribution
Conventional de-noising methods rely on the assumption that all samples are
independent and identically distributed, so the resultant classifier, though
disturbed by noise, can still easily identify the noises as the outliers of
training distribution. However, the assumption is unrealistic in large-scale
data that is inevitably long-tailed. Such imbalanced training data makes a
classifier less discriminative for the tail classes, whose previously "easy"
noises are now turned into "hard" ones -- they are almost as outliers as the
clean tail samples. We introduce this new challenge as Noisy Long-Tailed
Classification (NLT). Not surprisingly, we find that most de-noising methods
fail to identify the hard noises, resulting in significant performance drop on
the three proposed NLT benchmarks: ImageNet-NLT, Animal10-NLT, and Food101-NLT.
To this end, we design an iterative noisy learning framework called
Hard-to-Easy (H2E). Our bootstrapping philosophy is to first learn a classifier
as noise identifier invariant to the class and context distributional changes,
reducing "hard" noises to "easy" ones, whose removal further improves the
invariance. Experimental results show that our H2E outperforms state-of-the-art
de-noising methods and their ablations on long-tailed settings while
maintaining a stable performance on the conventional balanced settings.
Datasets and codes are available at https://github.com/yxymessi/H2E-FrameworkComment: Accepted to ECCV2022(Oral) ; Datasets and codes are available at
https://github.com/yxymessi/H2E-Framewor
- …